DELICIOUS: Deadline-Aware Approximate Computing in Cache-Conscious Multicore

نویسندگان

چکیده

Enhancing result-accuracy in approximate computing (AC) based real-time systems, without violating power constraints of the underlying hardware, is a challenging problem. Execution such AC applications can be split into two parts: (i) the mandatory part , execution which provides result acceptable quality, followed by (ii) optional that executed partially or fully to refine initially obtained order increase result-accuracy, time-constraint. This article introduces xmlns:xlink="http://www.w3.org/1999/xlink">DELICIOUS novel hybrid offline-online xmlns:xlink="http://www.w3.org/1999/xlink">scheduling strategy for dependent tasks. By employing an efficient xmlns:xlink="http://www.w3.org/1999/xlink">heuristic algorithm first generates schedule task-set with objective maximize results-accuracy, while respecting system-wide constraints. During execution, then xmlns:xlink="http://www.w3.org/1999/xlink">prudential cache resizing reduces temperature adjacent cores, generating thermal buffers at turned off ways. further trades this benefits enhancing processing speed cores stipulated duration, called xmlns:xlink="http://www.w3.org/1999/xlink">V/F Spiking budget core, shorten length reduced runtime exploited either enhance dynamically adjusting part, reduce enabling sleep mode cores. While surpassing prior art, offers 80% its scheduling strategy, enhanced 8.3% online, reducing peak 5.8°C on average, as shown benchmark evaluation 4-core multicore.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Doppelgänger: a cache for approximate computing

Modern processors contain large last level caches (LLCs) that consume substantial energy and area yet are imperative for high performance. Cache designs have improved dramatically by considering reference locality. Data values are also a source of optimization. Compression and deduplication exploit data values to use cache storage more efficiently resulting in smaller caches without sacrificing...

متن کامل

Moola: Multicore Cache Simulator

Chip multiprocessors have become the normative architecture for medium and high performance processors. These devices introduce new questions and research topics. One such topic is exploring the design space of a cachememory hierarchy that prevents the memory accesses from being a limiting factor on system performance. Simulation of system workloads is a widely accepted method for evaluating pr...

متن کامل

Resilience-Driven STT-RAM Cache Architecture for Approximate Computing

High-end manycore microprocessors exhibit large-sized caches (32MB – 128MB) that consume a significant amount of total energy. These caches are typically composed of 6T-SRAM cells, which lack efficiency in terms of area and leakage power [1][2]. The emerging memory technologies, like Spin-Transfer Torque RAM (STT-RAM), not only incur reduced leakage power but also provide high integration densi...

متن کامل

Cache-Conscious Query Processing

Foundations A cache is a hardware unit that speeds up access to data. Several cache units may be present at various levels of the memory hierarchy, depending on the processor architecture. For example, a processor may have a small but fast Level-1 (L1) cache for data, and another L1 cache for instructions. The same processor may have a larger but slower L2 cache storing both data and instructio...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Transactions on Parallel and Distributed Systems

سال: 2023

ISSN: ['1045-9219', '1558-2183', '2161-9883']

DOI: https://doi.org/10.1109/tpds.2022.3228751